Multimodal End-to-End Autonomous Driving
نویسندگان
چکیده
A crucial component of an autonomous vehicle (AV) is the artificial intelligence (AI) able to drive towards a desired destination. Today, there are different paradigms addressing development AI drivers. On one hand, we find modular pipelines, which divide driving task into sub-tasks such as perception and maneuver planning control. other end-to-end approaches that try learn direct mapping from input raw sensor data control signals. The later relatively less studied, but gaining popularity since they demanding in terms annotation. This paper focuses on driving. So far, most proposals relying this paradigm assume RGB images data. However, AVs will not be equipped only with cameras, also active sensors providing accurate depth information ( e.g. , LiDARs). Accordingly, analyses whether combining modalities, xmlns:xlink="http://www.w3.org/1999/xlink">i.e. using RGBD data, produces better drivers than single modality. We consider multimodality based early, mid late fusion schemes, both multisensory single-sensor (monocular estimation) settings. Using CARLA simulator conditional imitation learning (CIL), show how, indeed, early outperforms single-modality.
منابع مشابه
Query-Efficient Imitation Learning for End-to-End Autonomous Driving
One way to approach end-to-end autonomous driving is to learn a policy function that maps from a sensory input, such as an image frame from a front-facing camera, to a driving action, by imitating an expert driver, or a reference policy. This can be done by supervised learning, where a policy function is tuned to minimize the difference between the predicted and ground-truth actions. A policy f...
متن کاملLearning End-to-end Multimodal Sensor Policies for Autonomous Navigation
Sensor fusion is indispensable to improve accuracy and robustness in an autonomous navigation setting. However, in the space of end-to-end sensorimotor control, this multimodal outlook has received limited attention. In this work, we propose a novel stochastic regularization technique, called Sensor Dropout, to robustify multimodal sensor policy learning outcomes. We also introduce an auxiliary...
متن کاملDDD17: End-To-End DAVIS Driving Dataset
Event cameras, such as dynamic vision sensors (DVS), and dynamic and activepixel vision sensors (DAVIS) can supplement other autonomous driving sensors by providing a concurrent stream of standard active pixel sensor (APS) images and DVS temporal contrast events. The APS stream is a sequence of standard grayscale global-shutter image sensor frames. The DVS events represent brightness changes oc...
متن کاملAgile Off-Road Autonomous Driving Using End-to-End Deep Imitation Learning
We present an end-to-end imitation learning system for agile, off-road autonomous driving using only low-cost on-board sensors. By imitating a model predictive controller equipped with advanced sensors, we train a deep neural network control policy to map raw, high-dimensional observations to continuous steering and throttle commands. Compared with recent approaches to similar tasks, our method...
متن کاملEnd-to-end esophagojejunostomy versus standard end-to-side esophagojejunostomy: which one is preferable?
Abstract Background: End-to-side esophagojejunostomy has almost always been associated with some degree of dysphagia. To overcome this complication we decided to perform an end-to-end anastomosis and compare it with end-to-side Roux-en-Y esophagojejunostomy. Methods: In this prospective study, between 1998 and 2005, 71 patients with a diagnosis of gastric adenocarcinoma underwent total gastrec...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Intelligent Transportation Systems
سال: 2022
ISSN: ['1558-0016', '1524-9050']
DOI: https://doi.org/10.1109/tits.2020.3013234